white subject
Does a Rising Tide Lift All Boats? Bias Mitigation for AI-based CMR Segmentation
Lee, Tiarna, Puyol-Antón, Esther, Ruijsink, Bram, Shi, Miaojing, King, Andrew P.
Artificial intelligence (AI) is increasingly being used for medical imaging tasks. However, there can be biases in the resulting models, particularly when they were trained using imbalanced training datasets. One such example has been the strong race bias effect in cardiac magnetic resonance (CMR) image segmentation models. Although this phenomenon has been reported in a number of publications, little is known about the effectiveness of bias mitigation algorithms in this domain. We aim to investigate the impact of common bias mitigation methods to address bias between Black and White subjects in AI-based CMR segmentation models. Specifically, we use oversampling, importance reweighing and Group DRO as well as combinations of these techniques to mitigate the race bias. Furthermore, motivated by recent findings on the root causes of AI-based CMR segmentation bias, we evaluate the same methods using models trained and evaluated on cropped CMR images. We find that bias can be mitigated using oversampling, significantly improving performance for the underrepresented Black subjects whilst not significantly reducing the majority White subjects' performance. Group DRO also improves performance for Black subjects but not significantly, while reweighing decreases performance for Black subjects. Using a combination of oversampling and Group DRO also improves performance for Black subjects but not significantly. Using cropped images increases performance for both races and reduces the bias, whilst adding oversampling as a bias mitigation technique with cropped images reduces the bias further.
- North America > United States (0.14)
- Europe > United Kingdom > England > Greater London > London (0.05)
- North America > Canada > Alberta > Census Division No. 6 > Calgary Metropolitan Region > Calgary (0.04)
- Asia > China (0.04)
- Research Report > Experimental Study (0.70)
- Research Report > New Finding (0.69)
An investigation into the causes of race bias in AI-based cine CMR segmentation
Lee, Tiarna, Puyol-Anton, Esther, Ruijsink, Bram, Roujol, Sebastien, Barfoot, Theodore, Ogbomo-Harmitt, Shaheim, Shi, Miaojing, King, Andrew P.
Artificial intelligence (AI) methods are being used increasingly for the automated segmentation of cine cardiac magnetic resonance (CMR) imaging. However, these methods have been shown to be subject to race bias, i.e. they exhibit different levels of performance for different races depending on the (im)balance of the data used to train the AI model. In this paper we investigate the source of this bias, seeking to understand its root cause(s) so that it can be effectively mitigated. We perform a series of classification and segmentation experiments on short-axis cine CMR images acquired from Black and White subjects from the UK Biobank and apply AI interpretability methods to understand the results. In the classification experiments, we found that race can be predicted with high accuracy from the images alone, but less accurately from ground truth segmentations, suggesting that the distributional shift between races, which is often the cause of AI bias, is mostly image-based rather than segmentation-based. The interpretability methods showed that most attention in the classification models was focused on non-heart regions, such as subcutaneous fat. Cropping the images tightly around the heart reduced classification accuracy to around chance level. Similarly, race can be predicted from the latent representations of a biased segmentation model, suggesting that race information is encoded in the model. Cropping images tightly around the heart reduced but did not eliminate segmentation bias. We also investigate the influence of possible confounders on the bias observed.
- North America > United States (0.14)
- Europe > United Kingdom > England > Greater London > London (0.05)
- North America > Canada > Alberta > Census Division No. 6 > Calgary Metropolitan Region > Calgary (0.04)
- Asia > China (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Health & Medicine > Therapeutic Area > Cardiology/Vascular Diseases (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.49)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.47)
Is Facial Recognition Biased at Near-Infrared Spectrum As Well?
Krishnan, Anoop, Neas, Brian, Rattani, Ajita
Published academic research and media articles suggest face recognition is biased across demographics. Specifically, unequal performance is obtained for women, dark-skinned people, and older adults. However, these published studies have examined the bias of facial recognition in the visible spectrum (VIS). Factors such as facial makeup, facial hair, skin color, and illumination variation have been attributed to the bias of this technology at the VIS. The near-infrared (NIR) spectrum offers an advantage over the VIS in terms of robustness to factors such as illumination changes, facial makeup, and skin color. Therefore, it is worthwhile to investigate the bias of facial recognition at the near-infrared spectrum (NIR). This first study investigates the bias of the face recognition systems at the NIR spectrum. To this aim, two popular NIR facial image datasets namely, CASIA-Face-Africa and Notre-Dame-NIVL consisting of African and Caucasian subjects, respectively, are used to investigate the bias of facial recognition technology across gender and race. Interestingly, experimental results suggest equitable face recognition performance across gender and race at the NIR spectrum.
- Africa (0.26)
- South America > Brazil > São Paulo (0.04)
- North America > United States > Florida > Pinellas County (0.04)
- Research Report > New Finding (0.88)
- Research Report > Experimental Study (0.68)
A systematic study of race and sex bias in CNN-based cardiac MR segmentation
Lee, Tiarna, Puyol-Anton, Esther, Ruijsink, Bram, Shi, Miaojing, King, Andrew P.
In computer vision there has been significant research interest in assessing potential demographic bias in deep learning models. One of the main causes of such bias is imbalance in the training data. In medical imaging, where the potential impact of bias is arguably much greater, there has been less interest. In medical imaging pipelines, segmentation of structures of interest plays an important role in estimating clinical biomarkers that are subsequently used to inform patient management. Convolutional neural networks (CNNs) are starting to be used to automate this process. We present the first systematic study of the impact of training set imbalance on race and sex bias in CNN-based segmentation. We focus on segmentation of the structures of the heart from short axis cine cardiac magnetic resonance images, and train multiple CNN segmentation models with different levels of race/sex imbalance. We find no significant bias in the sex experiment but significant bias in two separate race experiments, highlighting the need to consider adequate representation of different demographic groups in health datasets.
- North America > United States (0.14)
- Europe > United Kingdom > England > Greater London > London (0.05)
- North America > Canada > Alberta > Census Division No. 6 > Calgary Metropolitan Region > Calgary (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.98)
- Health & Medicine > Therapeutic Area > Cardiology/Vascular Diseases (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Health & Medicine > Health Care Technology (0.88)
Microsoft is removing emotion recognition features from its facial recognition tech
When Microsoft announced last week it will remove several features from its facial recognition technology that deal with emotion, the head of its responsible artificial intelligence efforts included a warning: The science of emotion is far from settled. "Experts inside and outside the company have highlighted the lack of scientific consensus on the definition of'emotions,' the challenges in how inferences generalize across use cases, regions, and demographics, and the heightened privacy concerns around this type of capability," Natasha Crampton, Microsoft's chief responsible AI officer, wrote in a blog post. Microsoft's move, which came as part of a broader announcement about its "Responsible AI Standard" initiative, immediately became the most high-profile example of a company moving away from emotion recognition AI, a relatively small piece of technology that has been the focus of intense criticism, particularly in the academic community. Emotion recognition technology typically relies on software to look at any number of qualities -- facial expressions, tone of voice or word choice -- in an effort to automatically detect emotional state. Many technology companies have released software that claims to be able to read, recognize or measure emotions for use in business, education and customer service.
- North America > United States > Maryland (0.05)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.05)
Cohort Shapley value for algorithmic fairness
Mase, Masayoshi, Owen, Art B., Seiler, Benjamin B.
Cohort Shapley value is a model-free method of variable importance grounded in game theory that does not use any unobserved and potentially impossible feature combinations. We use it to evaluate algorithmic fairness, using the well known COMPAS recidivism data as our example. This approach allows one to identify for each individual in a data set the extent to which they were adversely or beneficially affected by their value of a protected attribute such as their race. The method can do this even if race was not one of the original predictors and even if it does not have access to a proprietary algorithm that has made the predictions. The grounding in game theory lets us define aggregate variable importance for a data set consistently with its per subject definitions. We can investigate variable importance for multiple quantities of interest in the fairness literature including false positive predictions.
- North America > United States > Florida > Broward County (0.04)
- North America > United States > Tennessee (0.04)
- North America > United States > New York (0.04)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- Research Report (1.00)
- Overview (0.93)
- Law (1.00)
- Leisure & Entertainment (0.68)